We consider optimization of generalized performance metrics for binaryclassification by means of surrogate losses. We focus on a class of metrics,which are linear-fractional functions of the false positive and false negativerates (examples of which include $F_{\beta}$-measure, Jaccard similaritycoefficient, AM measure, and many others). Our analysis concerns the followingtwo-step procedure. First, a real-valued function $f$ is learned by minimizinga surrogate loss for binary classification on the training sample. It isassumed that the surrogate loss is a strongly proper composite loss function(examples of which include logistic loss, squared-error loss, exponential loss,etc.). Then, given $f$, a threshold $\widehat{\theta}$ is tuned on a separatevalidation sample, by direct optimization of the target performance metric. Weshow that the regret of the resulting classifier (obtained from thresholding$f$ on $\widehat{\theta}$) measured with respect to the target metric isupperbounded by the regret of $f$ measured with respect to the surrogate loss.We also extend our results to cover multilabel classification and provideregret bounds for micro- and macro-averaging measures. Our findings are furtheranalyzed in a computational study on both synthetic and real data sets.
展开▼
机译:我们考虑通过代理损失优化针对二进制分类的广义性能指标。我们关注一类度量,它们是误报率和误报率的线性分数函数(示例包括$ F _ {\ beta} $度量,Jaccard相似系数,AM度量以及许多其他度量)。我们的分析涉及以下两步过程。首先,通过最小化训练样本的二进制分类的替代损失来学习实值函数$ f $。假设代理损失是一个很合适的复合损失函数(例如逻辑损失,平方误差损失,指数损失等)。然后,在给定$ f $的情况下,通过直接优化目标性能指标,在单独的验证样本上调整阈值$ \ widehat {\ theta} $。我们显示,针对目标指标衡量的所得分类器的后悔(从对$ \ widehat {\ theta} $的阈值$ f $获得)被针对代理损失的$ f $的遗憾所覆盖。将我们的结果扩展到涵盖多标签分类,并为微观和宏观平均度量提供遗憾的界限。我们的发现在合成和真实数据集的计算研究中得到了进一步分析。
展开▼